459 research outputs found
A Note on Minimax Testing and Confidence Intervals in Moment Inequality Models
This note uses a simple example to show how moment inequality models used in
the empirical economics literature lead to general minimax relative efficiency
comparisons. The main point is that such models involve inference on a low
dimensional parameter, which leads naturally to a definition of "distance"
that, in full generality, would be arbitrary in minimax testing problems. This
definition of distance is justified by the fact that it leads to a duality
between minimaxity of confidence intervals and tests, which does not hold for
other definitions of distance. Thus, the use of moment inequalities for
inference in a low dimensional parametric model places additional structure on
the testing problem, which leads to stronger conclusions regarding minimax
relative efficiency than would otherwise be possible
Optimal inference in a class of regression models
We consider the problem of constructing confidence intervals (CIs) for a
linear functional of a regression function, such as its value at a point, the
regression discontinuity parameter, or a regression coefficient in a linear or
partly linear regression. Our main assumption is that the regression function
is known to lie in a convex function class, which covers most smoothness and/or
shape assumptions used in econometrics. We derive finite-sample optimal CIs and
sharp efficiency bounds under normal errors with known variance. We show that
these results translate to uniform (over the function class) asymptotic results
when the error distribution is not known. When the function class is
centrosymmetric, these efficiency bounds imply that minimax CIs are close to
efficient at smooth regression functions. This implies, in particular, that it
is impossible to form CIs that are tighter using data-dependent tuning
parameters, and maintain coverage over the whole function class. We specialize
our results to inference on the regression discontinuity parameter, and
illustrate them in simulations and an empirical application.Comment: 39 pages plus supplementary material
Unbiased Instrumental Variables Estimation Under Known First-Stage Sign
We derive mean-unbiased estimators for the structural parameter in
instrumental variables models with a single endogenous regressor where the sign
of one or more first stage coefficients is known. In the case with a single
instrument, there is a unique non-randomized unbiased estimator based on the
reduced-form and first-stage regression estimates. For cases with multiple
instruments we propose a class of unbiased estimators and show that an
estimator within this class is efficient when the instruments are strong. We
show numerically that unbiasedness does not come at a cost of increased
dispersion in models with a single instrument: in this case the unbiased
estimator is less dispersed than the 2SLS estimator. Our finite-sample results
apply to normal models with known variance for the reduced-form errors, and
imply analogous results under weak instrument asymptotics with an unknown error
distribution
Adaptive Testing on a Regression Function at a Point
We consider the problem of inference on a regression function at a point when the entire function satisfies a sign or shape restriction under the null. We propose a test that achieves the optimal minimax rate adaptively over a range of Hölder classes, up to a log log n term, which we show to be necessary for adaptation. We apply the results to adaptive one-sided tests for the regression discontinuity parameter under a monotonicity restriction, the value of a monotone regression function at the boundary, and the proportion of true null hypotheses in a multiple testing problem
- …